MAPDP: Cooperative Multi-Agent Reinforcement Learning to Solve Pickup and Delivery Problems

نویسندگان

چکیده

Cooperative Pickup and Delivery Problem (PDP), as a variant of the typical Vehicle Routing Problems (VRP), is an important formulation in many real-world applications, such on-demand delivery, industrial warehousing, etc. It great importance to efficiently provide high-quality solutions cooperative PDP. However, it not trivial effective directly due two major challenges: 1) structural dependency between pickup delivery pairs require explicit modeling representation. 2) cooperation different vehicles highly related solution exploration difficult model. In this paper, we propose novel multi-agent reinforcement learning based framework solve PDP (MAPDP). First, design paired context embedding well measure nodes considering their limits. Second, utilize decoders leverage decision dependence among vehicle agents on special communication embedding. Third, A2C algorithm train integrated We conduct extensive experiments randomly generated dataset dataset. Experiments result shown that proposed MAPDP outperform all other baselines by at least 1.64\% settings, shows significant computation speed during inference.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Reinforcement Learning in Cooperative Multi–Agent Systems

Reinforcement Learning is used in cooperative multi–agent systems differently for various problems. We provide a review on learning algorithms used for repeated common–payoff games, and stochastic general– sum games. Then these learning algorithms is compared with another algorithm for the credit assignment problem that attempts to correctly assign agents the awards that they deserve.

متن کامل

Cooperative Multi-agent Control Using Deep Reinforcement Learning

This work considers the problem of learning cooperative policies in complex, partially observable domains without explicit communication. We extend three classes of single-agent deep reinforcement learning algorithms based on policy gradient, temporal-difference error, and actor-critic methods to cooperative multi-agent systems. We introduce a set of cooperative control tasks that includes task...

متن کامل

Argumentation Accelerated Reinforcement Learning for Cooperative Multi-Agent Systems

Multi-Agent Learning is a complex problem, especially in real-time systems. We address this problem by introducing Argumentation Accelerated Reinforcement Learning (AARL), which provides a methodology for defining heuristics, represented by arguments, and incorporates these heuristics into Reinforcement Learning (RL) by using reward shaping. We define AARL via argumentation and prove that it ca...

متن کامل

Cooperative Multi-Agent Reinforcement Learning for Low-Level Wireless Communication

Traditional radio systems are strictly co-designed on the lower levels of the OSI stack for compatibility and efficiency. Although this has enabled the success of radio communications, it has also introduced lengthy standardization processes and imposed static allocation of the radio spectrum. Various initiatives have been undertaken by the research community to tackle the problem of artificial...

متن کامل

Multi-Agent Reinforcement Learning: Independent vs. Cooperative Agents

Intelligent human agents exist in a cooperative social environment that facilitates learning. They learn not only by trial-and-error, but also through cooperation by sharing instantaneous information, episodic experience, and learned knowledge. The key investigations of this paper are, \Given the same number of reinforcement learning agents, will cooperative agents outperform independent agents...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2022

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v36i9.21236